Multiplicative Weights
نویسندگان
چکیده
In this lecture, we will study various applications of the theory of Multiplicative Weights (MW). In this section, we briefly review the general version of the MW algorithm that we studied in the previous lecture. The following sections then show how the theory can be applied to approximately solve zero-sum games and linear programs, and how it connects with the theory of boosting and approximation algorithms. We have n experts who predict the outcome of an event in consecutive rounds. Suppose that in each round there are P different outcomes for the event. If outcome j realizes, expert i pays a penalty of M(i, j). An important paremeter will prove to be the maximum allowable magnitude for the penalty. For that, let M(i, j) ∈ [−�, ρ], with 0 ≤ � ≤ ρ, where ρ is the width. Our goal is to devise a strategy that dictates which expert’s reccomendation to follow, in order to achieve an expected avegare penalty that is not much worse than that of the best expert (in hindsight). The strategy that we analyzed in the previous lecture is as follows. We maintain for each expert a scalar weight, which can be thought of as a quality score. Then, at each round we choose to follow the recommendation of a specific expert with probability that is proportional to her weight. After the outcome is realized, we update the weights of each expert accordingly. In mathematical terms, let wi t be the weight of the ith epxert at the beginning of round t. Then, the MW algorithm is
منابع مشابه
Performance of Multiplicative-weights-updates
Let us remember the mechanics of Multiplicative-Weights-Updates: At every time t, the learner maintains a weight vector wt ≥ 0 over the experts. Given the weight vector, the probability distribution over the experts is computed as pt = wt wt·1 . The weights are initialized at w1 = 1 n · 1. (Multiplicative-weights-update step.) Given the loss vector at time t the weights are updated as follows w...
متن کاملA Multiplicative Approach to Derive Weights in the Interval Analytic Hierarchy Process
This paper proposes a multiplicative approach to solve defective issues in lexicographic goal programming of the analytic hierarchy process (LGPAHP). Although the LGPAHP can handle inconsistent interval comparison matrices, it generates different weights in the upper and lower triangular interval judgments. Instead of adopting additive constraints, the proposed method uses multiplicative constr...
متن کاملCS261: A Second Course in Algorithms Lecture #12: Applications of Multiplicative Weights to Games and Linear Programs∗
1 Extensions of the Multiplicative Weights Guarantee Last lecture we introduced the multiplicative weights algorithm for online decision-making. You don't need to remember the algorithm details for this lecture, but you should remember that it's a simple and natural algorithm (just one simple update per action per time step). You should also remember its regret guarantee, which we proved last l...
متن کاملA Scale Mixture Perspective of Multiplicative Noise in Neural Networks
Corrupting the input and hidden layers of deep neural networks (DNNs) with multiplicative noise, often drawn from the Bernoulli distribution (or ‘dropout’), provides regularization that has significantly contributed to deep learning’s success. However, understanding how multiplicative corruptions prevent overfitting has been difficult due to the complexity of a DNN’s functional form. In this pa...
متن کاملEvolutionary stability implies asymptotic stability under multiplicative weights
We show that evolutionarily stable states in general (nonlinear) population games (which can be viewed as continuous vector fields constrained on a polytope) are asymptotically stable under a multiplicative weights dynamic (under appropriate choices of a parameter called the learning rate or step size, which we demonstrate to be crucial to achieve convergence, as otherwise even chaotic behavior...
متن کاملOrie 6334 Spectral Graph Theory Lecture 21
Just like matrix Chernoff bounds were a generalization of scalar Chernoff bounds, the multiplicative weights algorithm can be generalized to matrices. Recall that in the setup for the multiplicative weight update algorithm, we had a sequence of time steps t = 1, . . . , T ; in each time step t, we made a decision i ∈ {1...N} and got a value vt(i) ∈ [0, 1]. After we made a decision in time step ...
متن کامل